Shannon Entropy Versus Renyi Entropy from a Cryptographic Viewpoint

نویسنده

  • Maciej Skorski
چکیده

We provide a new inequality that links two important entropy notions: Shannon Entropy H1 and collision entropy H2. Our formula gives the worst possible amount of collision entropy in a probability distribution, when its Shannon Entropy is fixed. While in practice it is easier to evaluate Shannon entropy than other entropy notions, it is well known in folklore that it does not provide a good estimate of randomness quality from a cryptographic viewpoint, except very special settings. Our results and techniques put this in a quantitative form, allowing us to precisely answer the following questions: (a) How accurately does Shannon entropy estimate uniformity? Concretely, if the Shannon entropy of an n-bit source X is n− , where is a small number, can we conclude that X is close to uniform? This question is motivated by uniformity tests based on entropy estimators, like Maurer’s Universal Test. (b) How much randomness can we extract having high Shannon entropy? That is, if the Shannon entropy of an n-bit source X is n−O(1), how many almost uniform bits can we retrieve, at least? This question is motivated by the folklore upper bound O(log(n)). (c) Can we use high Shannon entropy for key derivation? More precisely, if we have an n-bit source X of Shannon entropy n − O(1), can we use it as a secure key for some applications, such as squaresecure applications? This is motivated by recent improvements in key derivation obtained by Barak et al. (CRYPTO’11) and Dodis et al. (TCC’14), which consider keys with some entropy deficiency. Our approach involves convex optimization techniques, which yield the shape of the ”worst” distribution, and the use of the Lambert W function, by which we resolve equations coming from Shannon Entropy constraints. We believe that it may be useful and of independent interests elsewhere, particularly for studying Shannon Entropy with constraints.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

The Rate of Entropy for Gaussian Processes

In this paper, we show that in order to obtain the Tsallis entropy rate for stochastic processes, we can use the limit of conditional entropy, as it was done for the case of Shannon and Renyi entropy rates. Using that we can obtain Tsallis entropy rate for stationary Gaussian processes. Finally, we derive the relation between Renyi, Shannon and Tsallis entropy rates for stationary Gaussian proc...

متن کامل

Shannon Entropy , Renyi Entropy , and Information

This memo contains proofs that the Shannon entropy is the limiting case of both the Renyi entropy and the Tsallis entropy, or information. These results are also confirmed experimentally. We conclude with some general observations on the utility of entropy measures. A brief summary of the origins of the concept of physical entropy are provided in an appendix.

متن کامل

Network Entropies of the Chinese Financial Market

Based on the data from the Chinese financial market, this paper focuses on analyzing three types of network entropies of the financial market, namely, Shannon, Renyi and Tsallis entropies. The findings suggest that Shannon entropy can reflect the volatility of the financial market, that Renyi and Tsallis entropies also have this function when their parameter has a positive value, and that Renyi...

متن کامل

Estimators in Cryptography

One of the main problems in cryptography is to give criteria to provide good comparators of cipher systems. The security of a cipher system must include the security of the algorithm, the security of the key generator and management module (see [BM94], [CM97],[Mau92a]) and the security of the cryptographic key agreement protocol (see [Mau93a],[MC94],[Mau93b],[Mau92b]). This paper gives show the...

متن کامل

The Rate of Entropy for Gaussian Processes

In this paper, we show that in order to obtain the Tsallis entropy rate for stochastic processes, we can use the limit of conditional entropy, as it was done for the case of Shannon and Renyi entropy rates. Using that we can obtain Tsallis entropy rate for stationary Gaussian processes. Finally, we derive the relation between Renyi, Shannon and Tsallis entropy rates for stationary Gaussian proc...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2015